MPI at Exascale
نویسندگان
چکیده
With petascale systems already available, researchers are devoting their attention to the issues needed to reach the next major level in performance, namely, exascale. Explicit message passing using the Message Passing Interface (MPI) is the most commonly used model for programming petascale systems today. In this paper, we investigate what is needed to enable MPI to scale to exascale, both in the MPI specification and in MPI implementations, focusing on issues such as memory consumption and performance. We also present results of experiments related to MPI memory consumption at scale on the IBM Blue Gene/P at Argonne National Laboratory.
منابع مشابه
CloverLeaf: Preparing Hydrodynamics Codes for Exascale
In this work we directly evaluate five candidate programming models for future exascale applications (MPI, MPI+OpenMP, MPI+OpenACC, MPI+CUDA and CAF) using a recently developed Lagrangian-Eulerian explicit hydrodynamics mini-application. The aim of this work is to better inform the exacsale planning at large HPC centres such as AWE. Such organisations invest significant resources maintaining an...
متن کاملScalable Memory Use in MPI: A Case Study with MPICH2
One of the factors that can limit the scalability of MPI to exascale is the amount of memory consumed by the MPI implementation. In fact, some researchers believe that existing MPI implementations, if used unchanged, will themselves consume a large fraction of the available system memory at exascale. To investigate and address this issue, we undertook a study of the memory consumed by the MPICH...
متن کاملAn Overview of MPI Characteristics of Exascale Proxy Applications
The scale of applications and computing systems is tremendously increasing and needs to increase even more to realize exascale systems. As the number of nodes keeps growing, communication has become key to high performance. The Message Passing Interface (MPI) has evolved to the de facto standard for inter-node data transfers. Consequently, MPI is well suited to serve as proxy for an analysis of...
متن کاملLightweight threading with MPI using Persistent Communications Semantics
Multi-threaded performance in MPI is of concern for future systems, particularly at Exascale, where massive concurrency will be necessary to leverage the full power of systems. While MPI provides generalized solutions and additional proposals like endpoints expand this general model, examining common use cases that have good solutions that may not be universally applicable is a viable additiona...
متن کاملSlouching Towards Exascale
One question before the high-performance computing community is "How will application developers write code for exascale machines?" At this point it looks like they might be riding a rough beast indeed. This talk is a brief assessment of where we stand now with respect to writing programs for our largest supercomputers and what we should do next. MPI is likely to remain a critical part of the p...
متن کامل